In the mathematics field of differential geometry, a metric tensor (or simply metric) is an additional structure on a manifold (such as a surface) that allows defining distances and angles, just as the inner product on a Euclidean space allows defining distances and angles there. More precisely, a metric tensor at a point of is a bilinear form defined on the tangent space at (that is, a bilinear function that maps pairs of to ), and a metric field on consists of a metric tensor at each point of that varies smoothly with .
A metric tensor is positive-definite if for every nonzero vector . A manifold equipped with a positive-definite metric tensor is known as a Riemannian manifold. Such a metric tensor can be thought of as specifying infinitesimal distance on the manifold. On a Riemannian manifold , the length of a smooth curve between two points and can be defined by integration, and the distance between and can be defined as the infimum of the lengths of all such curves; this makes a metric space. Conversely, the metric tensor itself is the derivative of the distance function (taken in a suitable manner).
While the notion of a metric tensor was known in some sense to mathematicians such as Gauss from the early 19th century, it was not until the early 20th century that its properties as a tensor were understood by, in particular, Gregorio Ricci-Curbastro and Tullio Levi-Civita, who first codified the notion of a tensor. The metric tensor is an example of a tensor field.
The components of a metric tensor in a coordinate basis take on the form of a symmetric matrix whose entries transform covariantly under changes to the coordinate system. Thus a metric tensor is a covariant symmetric tensor. From the Coordinate-free point of view, a metric tensor field is defined to be a nondegenerate symmetric bilinear form on each tangent space that varies smooth function from point to point.
depending on an ordered pair of real variables , and defined in an open set in the -plane. One of the chief aims of Gauss's investigations was to deduce those features of the surface which could be described by a function which would remain unchanged if the surface underwent a transformation in space (such as bending the surface without stretching it), or a change in the particular parametric form of the same geometrical surface.
One natural such invariant quantity is the arclength drawn along the surface. Another is the angle between a pair of curves drawn along the surface and meeting at a common point. A third such quantity is the area of a piece of the surface. The study of these invariants of a surface led Gauss to introduce the predecessor of the modern notion of the metric tensor.
The metric tensor is in the description below; E, F, and G in the matrix can contain any number as long as the matrix is positive definite.
s &= \int_a^b\left\|\frac{d}{dt}\vec{r}(u(t),v(t))\right\|\,dt \\[5pt] &= \int_a^b \sqrt{u'(t)^2\,\vec{r}_u\cdot\vec{r}_u + 2u'(t)v'(t)\, \vec{r}_u\cdot\vec{r}_v + v'(t)^2\,\vec{r}_v\cdot\vec{r}_v}\, dt \,,\end{align}
where represents the Euclidean norm. Here the chain rule has been applied, and the subscripts denote partial derivatives:
The integrand is the restrictionMore precisely, the integrand is the pullback of this differential to the curve. to the curve of the square root of the (quadratic form) differential
where
The quantity in () is called the line element, while is called the first fundamental form of . Intuitively, it represents the principal part of the square of the displacement undergone by when is increased by units, and is increased by units.
Using matrix notation, the first fundamental form becomes
\begin{bmatrix} du & dv \end{bmatrix} \begin{bmatrix} E & F \\ F & G \end{bmatrix} \begin{bmatrix} du \\ dv \end{bmatrix}
The chain rule relates , , and to , , and via the matrix equation
where the superscript T denotes the matrix transpose. The matrix with the coefficients , , and arranged in this way therefore transforms by the Jacobian matrix of the coordinate change
J = \begin{bmatrix} \frac{\partial u}{\partial u'} & \frac{\partial u}{\partial v'} \\ \frac{\partial v}{\partial u'} & \frac{\partial v}{\partial v'}\end{bmatrix}\,.
A matrix which transforms in this way is one kind of what is called a tensor. The matrix
with the transformation law () is known as the metric tensor of the surface.
first observed the significance of a system of coefficients , , and , that transformed in this way on passing from one system of coordinates to another. The upshot is that the first fundamental form () is ''invariant'' under changes in the coordinate system, and that this follows exclusively from the transformation properties of , , and . Indeed, by the chain rule,
\begin{bmatrix} \dfrac{\partial u}{\partial u'} & \dfrac{\partial u}{\partial v'} \\ \dfrac{\partial v}{\partial u'} & \dfrac{\partial v}{\partial v'} \end{bmatrix} \begin{bmatrix} du' \\ dv' \end{bmatrix}
so that
ds^2 &= \begin{bmatrix} du & dv \end{bmatrix} \begin{bmatrix} E & F \\ F & G \end{bmatrix} \begin{bmatrix} du \\ dv \end{bmatrix} \\[6pt] &= \begin{bmatrix} du' & dv' \end{bmatrix} \begin{bmatrix} \dfrac{\partial u}{\partial u'} & \dfrac{\partial u}{\partial v'} \\[6pt] \dfrac{\partial v}{\partial u'} & \dfrac{\partial v}{\partial v'} \end{bmatrix}^\mathsf{T} \begin{bmatrix} E & F \\ F & G \end{bmatrix} \begin{bmatrix} \dfrac{\partial u}{\partial u'} & \dfrac{\partial u}{\partial v'} \\[6pt] \dfrac{\partial v}{\partial u'} & \dfrac{\partial v}{\partial v'} \end{bmatrix} \begin{bmatrix} du' \\ dv' \end{bmatrix} \\[6pt] &= \begin{bmatrix} du' & dv' \end{bmatrix} \begin{bmatrix} E' & F' \\ F' & G' \end{bmatrix} \begin{bmatrix} du' \\ dv' \end{bmatrix}\\[6pt] &= (ds')^2 \,.\end{align}
for suitable real numbers and . If two tangent vectors are given:
\mathbf{a} &= a_1\vec{r}_u + a_2\vec{r}_v \\ \mathbf{b} &= b_1\vec{r}_u + b_2\vec{r}_v\end{align}
then using the bilinear form of the dot product,
\mathbf{a} \cdot \mathbf{b} &= a_1 b_1 \vec{r}_u\cdot\vec{r}_u + a_1b_2 \vec{r}_u\cdot\vec{r}_v + a_2b_1 \vec{r}_v\cdot\vec{r}_u + a_2 b_2 \vec{r}_v\cdot\vec{r}_v \\[8pt] &= a_1 b_1 E + a_1b_2 F + a_2b_1 F + a_2b_2G. \\[8pt] &= \begin{bmatrix} a_1 & a_2 \end{bmatrix} \begin{bmatrix} E & F \\ F & G \end{bmatrix} \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} \,.\end{align}
This is plainly a function of the four variables , , , and . It is more profitably viewed, however, as a function that takes a pair of arguments and which are vectors in the -plane. That is, put
This is a symmetric function in and , meaning that
It is also bilinear form, meaning that it is linear in each variable and separately. That is,
g\left(\lambda\mathbf{a} + \mu\mathbf{a}', \mathbf{b}\right) &= \lambda g(\mathbf{a}, \mathbf{b}) + \mu g\left(\mathbf{a}', \mathbf{b}\right),\quad\text{and} \\ g\left(\mathbf{a}, \lambda\mathbf{b} + \mu\mathbf{b}'\right) &= \lambda g(\mathbf{a}, \mathbf{b}) + \mu g\left(\mathbf{a}, \mathbf{b}'\right)\end{align}
for any vectors , , , and in the plane, and any real numbers and .
In particular, the length of a tangent vector is given by
and the angle between two vectors and is calculated by
where denotes the cross product, and the absolute value denotes the length of a vector in Euclidean space. By Lagrange's identity for the cross product, the integral can be written
&\iint_D \sqrt{\left(\vec{r}_u\cdot\vec{r}_u\right) \left(\vec{r}_v\cdot\vec{r}_v\right) - \left(\vec{r}_u\cdot\vec{r}_v\right)^2}\,du\,dv \\[5pt] ={} &\iint_D \sqrt{EG - F^2}\,du\,dv\\[5pt] ={} &\iint_D \sqrt{\det \begin{bmatrix} E & F \\ F & G \end{bmatrix}}\, du\, dv\end{align}
where is the determinant.
g_p(aU_p + bV_p, Y_p) &= ag_p(U_p, Y_p) + bg_p(V_p, Y_p) \,, \quad \text{and} \\ g_p(Y_p, aU_p + bV_p) &= ag_p(Y_p, U_p) + bg_p(Y_p, V_p) \,.\end{align}
A metric tensor field on assigns to each point of a metric tensor in the tangent space at in a way that varies smooth function with . More precisely, given any open set of manifold and any (smooth) and on , the real function is a smooth function of .
The functions form the entries of an symmetric matrix, . If
Denoting the matrix by and arranging the components of the vectors and into and ,
where T and T denote the matrix transpose of the vectors and , respectively. Under a change of basis of the form
for some invertible matrix , the matrix of components of the metric changes by as well. That is,
or, in terms of the entries of this matrix,
For this reason, the system of quantities is said to transform covariantly with respect to changes in the frame .
The metric has components relative to this frame given by
Relative to a new system of local coordinates, say
the metric tensor will determine a different matrix of coefficients,
This new system of functions is related to the original by means of the chain rule
so that
Or, in terms of the matrices and ,
where denotes the Jacobian matrix of the coordinate change.
If is positive for all non-zero , then the metric is positive-definite at . If the metric is positive-definite at every , then is called a Riemannian metric. More generally, if the quadratic forms have constant signature independent of , then the signature of is this signature, and is called a pseudo-Riemannian metric. If is connected space, then the signature of does not depend on .
By Sylvester's law of inertia, a basis of tangent vectors can be chosen locally so that the quadratic form diagonalizes in the following manner
for some between 1 and . Any two such expressions of (at the same point of ) will have the same number of positive signs. The signature of is the pair of integers , signifying that there are positive signs and negative signs in any such expression. Equivalently, the metric has signature if the matrix of the metric has positive and negative .
Certain metric signatures which arise frequently in applications are:
The inverse metric transforms contravariantly, or with respect to the inverse of the change of basis matrix . Whereas the metric itself provides a way to measure the length of (or angle between) vector fields, the inverse metric supplies a means of measuring the length of (or angle between) covector fields; that is, fields of linear functionals.
To see this, suppose that is a covector field. To wit, for each point , determines a function defined on tangent vectors at so that the following linearity condition holds for all tangent vectors and , and all real numbers and :
As varies, is assumed to be a smooth function in the sense that
is a smooth function of for any smooth vector field .
Any covector field has components in the basis of vector fields . These are determined by
Denote the row vector of these components by
Under a change of by a matrix , changes by the rule
That is, the row vector of components transforms as a covariant vector.
For a pair and of covector fields, define the inverse metric applied to these two covectors by
The resulting definition, although it involves the choice of basis , does not actually depend on in an essential way. Indeed, changing basis to gives
&\alpha[\mathbf{f}A] G[\mathbf{f}A]^{-1} \beta[\mathbf{f}A]^\mathsf{T} \\ ={} &\left(\alpha[\mathbf{f}]A\right) \left(A^{-1}G[\mathbf{f}]^{-1} \left(A^{-1}\right)^\mathsf{T}\right) \left(A^\mathsf{T}\beta[\mathbf{f}]^\mathsf{T}\right) \\ ={} &\alpha[\mathbf{f}] G[\mathbf{f}]^{-1} \beta[\mathbf{f}]^\mathsf{T}.\end{align}
So that the right-hand side of equation () is unaffected by changing the basis to any other basis whatsoever. Consequently, the equation may be assigned a meaning independently of the choice of basis. The entries of the matrix are denoted by , where the indices and have been raised to indicate the transformation law ().
for some uniquely determined smooth functions . Upon changing the basis by a nonsingular matrix , the coefficients change in such a way that equation () remains true. That is,
Consequently, . In other words, the components of a vector transform contravariantly (that is, inversely or in the opposite way) under a change of basis by the nonsingular matrix . The contravariance of the components of is notationally designated by placing the indices of in the upper position.
A frame also allows covectors to be expressed in terms of their components. For the basis of vector fields define the dual basis to be the linear functionals such that
That is, , the Kronecker delta. Let
Under a change of basis for a nonsingular matrix , transforms via
Any linear functional on tangent vectors can be expanded in terms of the dual basis
where denotes the row vector . The components transform when the basis is replaced by in such a way that equation () continues to hold. That is,
whence, because , it follows that . That is, the components transform covariantly (by the matrix rather than its inverse). The covariance of the components of is notationally designated by placing the indices of in the lower position.
Now, the metric tensor gives a means to identify vectors and covectors as follows. Holding fixed, the function
of tangent vector defines a linear functional on the tangent space at . This operation takes a vector at a point and produces a covector . In a basis of vector fields , if a vector field has components , then the components of the covector field in the dual basis are given by the entries of the row vector
Under a change of basis , the right-hand side of this equation transforms via
v[\mathbf{f}A]^\mathsf{T} G[\mathbf{f}A] = v[\mathbf{f}]^\mathsf{T} \left(A^{-1}\right)^\mathsf{T} A^\mathsf{T} G[\mathbf{f}]A = v[\mathbf{f}]^\mathsf{T} G[\mathbf{f}]A
so that : transforms covariantly. The operation of associating to the (contravariant) components of a vector field T the (covariant) components of the covector field , where
is called lowering the index.
To raise the index, one applies the same construction but with the inverse metric instead of the metric. If are the components of a covector in the dual basis , then the column vector
has components which transform contravariantly:
Consequently, the quantity does not depend on the choice of basis in an essential way, and thus defines a vector field on . The operation () associating to the (covariant) components of a covector the (contravariant) components of a vector given is called raising the index. In components, () is
Suppose that is an immersion onto the submanifold . The usual Euclidean dot product in is a metric which, when restricted to vectors tangent to , gives a means for taking the dot product of these tangent vectors. This is called the induced metric.
Suppose that is a tangent vector at a point of , say
where are the standard coordinate vectors in . When is applied to , the vector goes over to the vector tangent to given by
(This is called the pushforward of along .) Given two such vectors, and , the induced metric is defined by
It follows from a straightforward calculation that the matrix of the induced metric in the basis of coordinate vector fields is given by
where is the Jacobian matrix:
\frac{\partial\varphi^1}{\partial x^1} & \frac{\partial\varphi^1}{\partial x^2} & \dots & \frac{\partial\varphi^1}{\partial x^n} \\[1ex] \frac{\partial\varphi^2}{\partial x^1} & \frac{\partial\varphi^2}{\partial x^2} & \dots & \frac{\partial\varphi^2}{\partial x^n} \\ \vdots & \vdots & \ddots & \vdots \\ \frac{\partial\varphi^m}{\partial x^1} & \frac{\partial\varphi^m}{\partial x^2} & \dots & \frac{\partial\varphi^m}{\partial x^n}\end{bmatrix}.
from the fiber product of the tangent bundle of with itself to such that the restriction of to each fiber is a nondegenerate bilinear mapping
The mapping () is required to be continuous, and often continuously differentiable, smooth function, or real analytic, depending on the case of interest, and whether can support such a structure.
The section is defined on simple elements of by
and is defined on arbitrary elements of by extending linearly to linear combinations of simple elements. The original bilinear form is symmetric if and only if
Since is finite-dimensional, there is a natural isomorphism
so that is regarded also as a section of the bundle of the cotangent bundle with itself. Since is symmetric as a bilinear mapping, it follows that is a symmetric tensor.
from the fiber product of to which is bilinear in each fiber:
Using duality as above, a metric is often identified with a section of the tensor product bundle .
the linear functional on which sends a tangent vector at to . That is, in terms of the pairing between and its dual space ,
for all tangent vectors and . The mapping is a linear transformation from to . It follows from the definition of non-degeneracy that the kernel of is reduced to zero, and so by the rank–nullity theorem, is a linear isomorphism. Furthermore, is a symmetric linear transformation in the sense that
for all tangent vectors and .
Conversely, any linear isomorphism defines a non-degenerate bilinear form on by means of
This bilinear form is symmetric if and only if is symmetric. There is thus a natural one-to-one correspondence between symmetric bilinear forms on and symmetric linear isomorphisms of to the dual .
As varies over , defines a section of the bundle of vector bundle isomorphisms of the tangent bundle to the cotangent bundle. This section has the same smoothness as : it is continuous, differentiable, smooth, or real-analytic according as . The mapping , which associates to every vector field on a covector field on gives an abstract formulation of "lowering the index" on a vector field. The inverse of is a mapping which, analogously, gives an abstract formulation of "raising the index" on a covector field.
The inverse defines a linear mapping
which is nonsingular and symmetric in the sense that
for all covectors , . Such a nonsingular symmetric mapping gives rise (by the tensor-hom adjunction) to a map
or by the Double dual to a section of the tensor product
Let be a piecewise-differentiable parametric curve in , for . The arclength of the curve is defined by
In connection with this geometrical application, the quadratic form differential form
is called the first fundamental form associated to the metric, while is the line element. When is pulled back to the image of a curve in , it represents the square of the differential with respect to arclength.
For a pseudo-Riemannian metric, the length formula above is not always defined, because the term under the square root may become negative. We generally only define the length of a curve when the quantity under the square root is always of one sign or the other. In this case, define
While these formulas use coordinate expressions, they are in fact independent of the coordinates chosen; they depend only on the metric, and the curve along which the formula is integrated.
This usage comes from physics, specifically, classical mechanics, where the integral can be seen to directly correspond to the kinetic energy of a point particle moving on the surface of a manifold. Thus, for example, in Jacobi's formulation of Maupertuis' principle, the metric tensor can be seen to correspond to the mass tensor of a moving particle.
In many cases, whenever a calculation calls for the length to be used, a similar calculation using the energy may be done as well. This often leads to simpler formulas by avoiding the need for the square-root. Thus, for example, the geodesic equations may be obtained by applying variational principles to either the length or the energy. In the latter case, the geodesic equations are seen to arise from the principle of least action: they describe the motion of a "free particle" (a particle feeling no forces) that is confined to move on the manifold, but otherwise moves freely, with constant momentum, within the manifold.
A measure can be defined, by the Riesz representation theorem, by giving a positive linear functional on the space of compact support continuous functions on . More precisely, if is a manifold with a (pseudo-)Riemannian metric tensor , then there is a unique positive Borel measure such that for any coordinate chart , for all supported in . Here is the determinant of the matrix formed by the components of the metric tensor in the coordinate chart. That is well-defined on functions supported in coordinate neighborhoods is justified by Jacobian change of variables. It extends to a unique positive linear functional on by means of a partition of unity.
If is also oriented, then it is possible to define a natural volume form from the metric tensor. In a positively oriented coordinate system the volume form is represented as where the are the coordinate differentials and denotes the exterior product in the algebra of differential forms. The volume form also gives a way to integrate functions on the manifold, and this geometric integral agrees with the integral obtained by the canonical Borel measure.
The length of a curve reduces to the formula:
The Euclidean metric in some other common coordinate systems can be written as follows.
Polar coordinates :
x &= r \cos\theta \\ y &= r \sin\theta \\ J &= \begin{bmatrix}\cos\theta & -r\sin\theta \\ \sin\theta & r\cos\theta\end{bmatrix} \,.\end{align}
So
\begin{bmatrix} \cos^2\theta + \sin^2\theta & -r\sin\theta \cos\theta + r\sin\theta\cos\theta \\ -r\cos\theta\sin\theta + r\cos\theta\sin\theta & r^2 \sin^2\theta + r^2\cos^2\theta \end{bmatrix} = \begin{bmatrix} 1 & 0 \\ 0 & r^2 \end{bmatrix}by trigonometric identities.
In general, in a Cartesian coordinate system on a Euclidean space, the partial derivatives are orthonormal with respect to the Euclidean metric. Thus the metric tensor is the Kronecker delta δ ij in this coordinate system. The metric tensor with respect to arbitrary (possibly curvilinear) coordinates is given by
\sum_{kl}\delta_{kl}\frac{\partial x^k}{\partial q^i} \frac{\partial x^l}{\partial q^j} = \sum_k\frac{\partial x^k}{\partial q^i}\frac{\partial x^k}{\partial q^j}.
This is usually written in the form
the metric is, depending on choice of metric signature,
For a curve with—for example—constant time coordinate, the length formula with this metric reduces to the usual length formula. For a timelike curve, the length formula gives the proper time along the curve.
In this case, the spacetime interval is written as
The Schwarzschild metric describes the spacetime around a spherically symmetric body, such as a planet, or a black hole. With coordinates
we can write the metric as
\begin{bmatrix} \left(1 - \frac{2GM}{rc^2}\right) & 0 & 0 & 0 \\ 0 & -\left(1 - \frac{2GM}{r c^2}\right)^{-1} & 0 & 0 \\ 0 & 0 & -r^2 & 0 \\ 0 & 0 & 0 & -r^2 \sin^2 \theta \end{bmatrix}\,,
where (inside the matrix) is the gravitational constant and represents the total mass–energy content of the central object.
|
|